Generative image captioning in Urdu using deep learning

نویسندگان

چکیده

Abstract Urdu is morphologically rich language and lacks the resources available in English. While several studies on image captioning task English have been published, this among pioneer generative captioning. The study makes key contributions: (i) it presents a new dataset for captioning, (ii) different attention-based architectures language. These attention mechanisms are to language, as those never used (iii) Finally, performs quantitative qualitative analysis of results by studying impact model Urdu’s caption generation task. extensive experiments show encouraging such BLEU-1 score 72.5, BLEU-2 56.9, BLEU-3 42.8, BLEU-4 31.6. we present data code future research via GitHub ( https://github.com/saeedhas/Urdu_cap_gen ).

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Deep Learning for Automatic Image Captioning in Poor Training Conditions

English. Recent advancements in Deep Learning show that the combination of Convolutional Neural Networks and Recurrent Neural Networks enables the definition of very effective methods for the automatic captioning of images. Unfortunately, this straightforward result requires the existence of large-scale corpora and they are not available for many languages. This paper describes a simple methodo...

متن کامل

Generative Choreography using Deep Learning

Recent advances in deep learning have enabled the extraction of high-level features from raw sensor data which has opened up new possibilities in many different fields, including computer generated choreography. In this paper we present a system chorrnn for generating novel choreographic material in the nuanced choreographic language and style of an individual choreographer. It also shows promi...

متن کامل

Learning to Evaluate Image Captioning

Evaluation metrics for image captioning face two challenges. Firstly, commonly used metrics such as CIDEr, METEOR, ROUGE and BLEU often do not correlate well with human judgments. Secondly, each metric has well known blind spots to pathological caption constructions, and rulebased metrics lack provisions to repair such blind spots once identified. For example, the newly proposed SPICE correlate...

متن کامل

Contrastive Learning for Image Captioning

Image captioning, a popular topic in computer vision, has achieved substantial progress in recent years. However, the distinctiveness of natural descriptions is often overlooked in previous work. It is closely related to the quality of captions, as distinctive captions are more likely to describe images with their unique aspects. In this work, we propose a new learning method, Contrastive Learn...

متن کامل

Deep Learning for Video Classification and Captioning

Accelerated by the tremendous increase in Internet bandwidth and storage space, video data has been generated, published and spread explosively, becoming an indispensable part of today's big data. In this paper, we focus on reviewing two lines of research aiming to stimulate the comprehension of videos with deep learning: video classification and video captioning. While video classification con...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of Ambient Intelligence and Humanized Computing

سال: 2023

ISSN: ['1868-5137', '1868-5145']

DOI: https://doi.org/10.1007/s12652-023-04584-y